Learning high-dimensional DAGs with latent and selection variables (Abstract)
نویسندگان
چکیده
We consider the problem of learning causal information between random variables in directed acyclic graphs (DAGs) when allowing arbitrarily many latent and selection variables. The Fast Causal Inference algorithm (FCI) (Spirtes et al., 1999) has been explicitly designed to infer conditional independence and causal information in such settings. Despite its name, FCI is computationally very intensive for large graphs. Spirtes (2001) introduced a modified version of FCI, called Anytime FCI, which only performs conditional independence tests up to a pre-specified cutoff k. Anytime FCI is typically faster but less informative than FCI, but the causal interpretation of tails and arrowheads in its output is still sound. We propose an adaptation of Anytime FCI, called Adaptive Anytime FCI (AAFCI), where the cut-off k is set to the maximum size of the conditioning sets used to find the initial skeleton in FCI. Moreover, we propose a new algorithm, called Really Fast Causal Inference (RFCI), which has similar properties as AAFCI but is much faster for large sparse graphs. The complete paper is available at http://arxiv.org/abs/1104.5617.
منابع مشابه
Learning high-dimensional directed acyclic graphs with latent and selection variables
We consider the problem of learning causal information between random variables in directed acyclic graphs (DAGs) when allowing arbitrarily many latent and selection variables. The FCI (Fast Causal Inference) algorithm has been explicitly designed to infer conditional independence and causal information in such settings. However, FCI is computationally infeasible for large graphs. We therefore ...
متن کاملA Polynomial Time Algorithm For Determining DAG Equivalence in the Presence of Latent Variables and Selection Bias
Following the terminology of Lauritzen et. al. (1990) say that a probability measure over a set of variables V satisfies the local directed Markov property for a directed acyclic graph (DAG) G with vertices V if and only if for every W in V, W is independent of the set of all its non-descendants conditional on the set of its parents. One natural question that arises with respect to DAGs is when...
متن کاملThe Hidden Life of Latent Variables: Bayesian Learning with Mixed Graph Models
Directed acyclic graphs (DAGs) have been widely used as a representation of conditional independence in machine learning and statistics. Moreover, hidden or latent variables are often an important component of graphical models. However, DAG models suffer from an important limitation: the family of DAGs is not closed under marginalization of hidden variables. This means that in general we cannot...
متن کاملLearning Linear Bayesian Networks with Latent Variables
This work considers the problem of learning linear Bayesian networks when some of the variables are unobserved. Identifiability and efficient recovery from low-order observable moments are established under a novel graphical constraint. The constraint concerns the expansion properties of the underlying directed acyclic graph (DAG) between observed and unobserved variables in the network, and it...
متن کاملA Transformational Characterization of Markov Equivalence for Directed Acyclic Graphs with Latent Variables
Different directed acyclic graphs (DAGs) may be Markov equivalent in the sense that they entail the same conditional independence relations among the observed variables. Chickering (1995) provided a transformational characterization of Markov equivalence for DAGs (with no latent variables), which is useful in deriving properties shared by Markov equivalent DAGs, and, with certain generalization...
متن کامل